Smaller generalization error derived for a deep residual neural network compared with shallow networks
نویسندگان
چکیده
Abstract Estimates of the generalization error are proved for a residual neural network with $L$ random Fourier features layers $\bar z_{\ell +1}=\bar z_\ell + \textrm {Re}\sum _{k=1}^K\bar b_{\ell k}\,e^{\textrm {i}\omega _{\ell k}\bar }+ c_{\ell ^{\prime}_{\ell k}\cdot x}$. An optimal distribution frequencies $(\omega k},\omega k})$ $e^{\textrm }$ and x}$ is derived. This derivation based on corresponding approximation function values $f(x)$. The turns out to be smaller than estimate ${\|\hat f\|^2_{L^1({\mathbb {R}}^d)}}/{(KL)}$ features, one hidden layer same total number nodes $KL$, in case $L^\infty $-norm $f$ much less $L^1$-norm its transform $\hat f$. understanding an used construct new training method deep network. Promising performance proposed algorithm demonstrated computational experiments.
منابع مشابه
Inferring Functional Neural Connectivity with Deep Residual Convolutional Networks
Measuring synaptic connectivity in large neuronal populations remains a major goal of modern neuroscience. While this connectivity is traditionally revealed by anatomical methods such as electron microscopy, an efficient alternative is to computationally infer functional connectivity from recordings of neural activity. However, these statistical techniques still require further refinement befor...
متن کاملGoGoGo: Improving Deep Neural Network Based Go Playing AI with Residual Networks
The game of Go has a long history and has been viewed as the most challanging classical game due to the enormous amount of possible moves and the lack of precise evaulation tools. AlphaGo [13], a Go-playing AI built by Google DeepMind, used a new approach of combining deep neural networks with tree search to solve the Go playing problem. It narrows down the search space by introducing ”policy n...
متن کاملReversible Architectures for Arbitrarily Deep Residual Neural Networks
Recently, deep residual networks have been successfully applied in many computer vision and natural language processing tasks, pushing the state-of-the-art performance with deeper and wider architectures. In this work, we interpret deep residual networks as ordinary differential equations (ODEs), which have long been studied in mathematics and physics with rich theoretical and empirical success...
متن کاملOn the complexity of shallow and deep neural network classifiers
Recently, deep networks were proved to be more effective than shallow architectures to face complex real–world applications. However, theoretical results supporting this claim are still few and incomplete. In this paper, we propose a new topological measure to study how the depth of feedforward networks impacts on their ability of implementing high complexity functions. Upper and lower bounds o...
متن کاملGeneralization Tower Network: A Novel Deep Neural Network Architecture for Multi-Task Learning
Deep learning (DL) advances state-of-the-art reinforcement learning (RL), by incorporating deep neural networks in learning representations from the input to RL. However, the conventional deep neural network architecture is limited in learning representations for multi-task RL (MT-RL), as multiple tasks can refer to different kinds of representations. In this paper, we thus propose a novel deep...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Ima Journal of Numerical Analysis
سال: 2022
ISSN: ['1464-3642', '0272-4979']
DOI: https://doi.org/10.1093/imanum/drac049